Discriminatively trained dependency language modeling for conversational speech recognition

نویسندگان

  • Benjamin Lambert
  • Bhiksha Raj
  • Rita Singh
چکیده

We present a discriminatively trained dependency parserbased language model. The model operates on utterances, rather than words, and so can utilize long-distance structural features of each sentence. We train the model discriminatively on n-best lists, using the perceptron algorithm to tune the model weights. Our features include standard n-gram style features, long-distance co-occurrence features, and syntactic structural features. We evaluate this model by re-ranking nbest lists of recognized speech from the Fisher dataset of informal telephone conversations. We compare various combinations of feature types, and methods of training the model.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Improved Single System Conversational Telephone Speech Recognition with VGG Bottleneck Features

On small datasets, discriminatively trained bottleneck features from deep networks commonly outperform more traditional spectral or cepstral features. While these features are typically trained with small, fully-connected networks, recent studies have used more sophisticated networks with great success. We use the recent deep CNN (VGG) network for bottleneck feature extraction—previously used o...

متن کامل

A language model for conversational speech recognition using information designed for speech translation

In this paper, a new language model is proposed for speech recognition in conversational speech translation. In conversation, speech strongly depends on the previous utterance of the other participant. Applying this dependency in language modeling, we can reduce the speech recognition error rate. To this end, we propose the following new language model where the content of the previous utteranc...

متن کامل

Discriminative training of WFST factors with application to pronunciation modeling

One of the most popular speech recognition architectures consists of multiple components (like the acoustic, pronunciation and language models) that are modeled as weighted finite state transducer (WFST) factors in a cascade. These factor WFSTs are typically trained in isolation and combined efficiently for decoding. Recent work has explored jointly estimating parameters for these models using ...

متن کامل

Pseudo-Syntactic Language Modeling for Disfluent Speech Recognition

Abstract Language models for speech recognition are generally trained on text corpora. Since these corpora do not contain the disfluencies found in natural speech, there is a train/test mismatch when these models are applied to conversational speech. In this work we investigate a language model (LM) designed to model these disfluencies as a syntactic process. By modeling selfcorrections we obta...

متن کامل

Discriminative training of stochastic Markov graphs for speech recognition

This paper proposes the application of discriminative training techniques based on the Generalized Probabilistic Descent (GPD) approach to Stochastic Markov Graphs (SMGs), a generalization of mixture-state Hidden Markov Models (HMMs), describing the constraints in the acoustic structure of speech as a graph consisting of nodes, each containing a base function, and a transition network between t...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2013